Minimizing System Correlation in SVM Training

نویسنده

  • Luciana Ferrer
چکیده

1 Description of the problem We will consider a binary classiication task for which two separate classiiers are available. Each classiier may use diierent input features and diierent modeling techniques. In a setup like this, the nal decision is made based on a combination of the outputs generated by both classiiers with the hope that the nal performance will be better than the performance of the two individual classiiers. This, nevertheless, is not necessarily the case. In the extreme, if both classiiers were generating exactly the same output for each sample, the combined classiier could never have a better performance than the individual ones, independently of the combination procedure used. Intuitively, what we wish is to have two classiiers for which the within class correlation is small. This way, both classiiers contribute independent information leading to a better nal decision. In this work we will study the case in which one of the classiiers is given to us (we can consider this system as a black box which simply gives us a value for each sample) and the other one is an SVM which we need to train. Our goal is to modify the training criteria for the SVM so that the score resulting from this system is as little correlated as possible to the scores from the black box system. 2 Anti-correlation Kernel In this section we will derive the optimization problem we need to solve in order to achieve the combined goal of minimizing the error of the SVM system (which we will call S) while also minimizing the correlation of this system with the original black box system (which we will call B). Given a training set T = f(x We want to modify the objective function by adding a term 2 in the objective function, where is a tunable parameter and is the within-class correlation between system S and system B. Given the scores fb (i) ; i = 1; :::; mg from system B for the training set T, we can compute the within-class correlation between the scores produced by the SVM and these scores the following way: 2 = cov(B; SjY) 2 var(BjY)var(SjY) (2) where cov(B; SjY), var(BjY) and var(SjY) are the within-class covariance and variances. These can be approximated by the within-class sample covariance and variances in the training set T. The within-class sample covariance can be calculated as,

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

An Efficient Method for Gradient-Based Adaptation of Hyperparameters in SVM Models

We consider the task of tuning hyperparameters in SVM models based on minimizing a smooth performance validation function, e.g., smoothed k-fold crossvalidation error, using non-linear optimization techniques. The key computation in this approach is that of the gradient of the validation function with respect to hyperparameters. We show that for large-scale problems involving a wide choice of k...

متن کامل

A New Formulation for Cost-Sensitive Two Group Support Vector Machine with Multiple Error Rate

Support vector machine (SVM) is a popular classification technique which classifies data using a max-margin separator hyperplane. The normal vector and bias of the mentioned hyperplane is determined by solving a quadratic model implies that SVM training confronts by an optimization problem. Among of the extensions of SVM, cost-sensitive scheme refers to a model with multiple costs which conside...

متن کامل

Classification of transformer faults using frequency response analysis based on cross-correlation technique and support vector machine

One of the most important methods for transformers fault diagnosis (especially mechanical defects) is the frequency response analysis (FRA) method. The most important step in the FRA diagnostic process is to differentiate the faults and classify them in different classes. This paper uses the intelligent support vector machine (SVM) method to classify transformer faults. For this purpose, two gr...

متن کامل

Properties of Support Vector Machines for Regression Properties of Support Vector Machines for Regression

In this report we show that the-tube size in Support Vector Machine (SVM) for regression is 2= p 1 + jjwjj 2. By using this result we show that, in the case all the data points are inside the-tube, minimizing jjwjj 2 in SVM for regression is equivalent to maximizing the distance between the approximating hyperplane and the farest points in the training set. Moreover, in the most general setting...

متن کامل

Hedging of Options in Jump-Diffusion Markets with Correlated Assets

We consider the hedging problem in a jump-diffusion market with correlated assets. For this purpose, we employ the locally risk-minimizing approach and obtain the hedging portfolio as a solution of a multidimensional system of linear equations. ‎This system shows that in a continuous market, independence and correlation assumptions of assets lead to the same locally risk-minimizing portfolio. ‎...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2006